Optimization of the Error Entropy Minimization Algorithm for Neural Network Classification
نویسنده
چکیده
One way of using entropy criteria in learning systems is to minimize the entropy of the error between the output of the learning system and the desired targets. In our last work, we introduced the Error Entropy Minimization (EEM) algorithm for neural network classification. There are some sensible aspects in the optimization of the EEM algorithm: the size of the Parzen Window (smoothing parameter) and the value of the learning rate parameter are the two most important. In this paper, we show how to optimize the EEM algorithm by using a variable learning rate during the training phase. INTRODUCTION The Error Entropy Minimization (EEM) algorithm minimizes the Reniy’s Quadratic Entropy of the error between the output of the neural network and the desired targets. Reniy’s Quadratic Entropy is used because, when using a nonparametric estimate of the probability density function (pdf) by the Parzen Window method, a simplified expression for the entropy is obtained. The size of the Kernel window, h, used in Parzen Window method is one of the two most sensible aspects in the implementation of the EEM algorithm. The other important aspect is the backpropagation learning rate. We implemented a variable learning rate, similar to the one used in the MSE algorithm, and the results were better than the ones obtained with the previous algorithm. RENYI'S QUADRATIC ENTROPY AND BACK-PROPAGATION ALGORITHM Renyi extended the concept of entropy introduced by Shanon and defined the Renyi's α entropy, applied to continuos random variables, as (Renyi 1976): [ ] ∫ − = C R dx x f H α α ) ( log (1) Renyi's Quadratic Entropy, 2 = α , in conjunction with the Parzen Window pdf estimation with gaussian kernel allows, the determination of entropy in a non-parametric, practical and computationally efficient way. The Parzen window method estimates the pdf ) (x f as ∑ = ⎟ ⎠ ⎞ ⎜ ⎝ ⎛ − = N i i m h x x K Nh x f 1 1 ) ( (2) where N is the number of data points, K is a kernel function, m is the dimensionality of vectors x ( m x R ∈ ) and h the bandwidth or smoothing parameter. We usualy use the simplest Gaussian kernel with zero mean and Σ equal to I (the identity matrix) ⎟ ⎠ ⎞ ⎜ ⎝ ⎛− = x x x G T m 2 1 exp
منابع مشابه
A New Method for Intrusion Detection Using Genetic Algorithm and Neural Network
The article attempts to have neural network and genetic algorithm techniques present a model for classification on dataset. The goal is design model can the subject acted a firewall in network and this model with compound optimized algorithms create reliability and accuracy and reduce error rate couse of this is article use feedback neural network and compared to previous methods increase a...
متن کاملA New Method for Intrusion Detection Using Genetic Algorithm and Neural Network
The article attempts to have neural network and genetic algorithm techniques present a model for classification on dataset. The goal is design model can the subject acted a firewall in network and this model with compound optimized algorithms create reliability and accuracy and reduce error rate couse of this is article use feedback neural network and compared to previous methods increase a...
متن کاملA Neural Network Method Based on Mittag-Leffler Function for Solving a Class of Fractional Optimal Control Problems
In this paper, a computational intelligence method is used for the solution of fractional optimal control problems (FOCP)'s with equality and inequality constraints. According to the Ponteryagin minimum principle (PMP) for FOCP with fractional derivative in the Riemann- Liouville sense and by constructing a suitable error function, we define an unconstrained minimization problem. In the optimiz...
متن کاملNeural network classification using Shannon's entropy
The last years have witnessed an increasing attention to entropy-based criteria in adaptive systems. Several principles were proposed based on the maximization or minimization of entropic cost functions. We propose a new type of neural network classifiers with multilayer perceptron (MLP) architecture, but where the usual mean square error minimization principle is substituted by the minimizatio...
متن کاملThe Error Entropy Minimization Algorithm for Neural Network Classification
One way of using the entropy criteria in learning systems is to minimize the entropy of the error between two variables: typically, one is the output of the learning system and the other is the target. This framework has been used for regression. In this paper we show how to use the minimization of the entropy of the error for classification. The minimization of the entropy of the error implies...
متن کامل